809 research outputs found

    Resilience to COVID-19: Socioeconomic Disadvantage Associated With Positive Caregiver-Youth Communication and Youth Preventative Actions

    Get PDF
    Socioeconomic disadvantage is associated with larger COVID-19 disease burdens and pandemic-related economic impacts. We utilized the longitudinal Adolescent Brain Cognitive Development Study to understand how family- and neighborhood-level socioeconomic disadvantage relate to disease burden, family communication, and preventative responses to the pandemic in over 6,000 youth-caregiver dyads. Data were collected at three timepoints (May-August 2020). Here, we show that both family- and neighborhood-level disadvantage were associated with caregivers\u27 reports of greater family COVID-19 disease burden, less perceived exposure risk, more frequent caregiver-youth conversations about COVID-19 risk/prevention and reassurance, and greater youth preventative behaviors. Families with more socioeconomic disadvantage may be adaptively incorporating more protective strategies to reduce emotional distress and likelihood of COVID-19 infection. The results highlight the importance of caregiver-youth communication and disease-preventative practices for buffering the economic and disease burdens of COVID-19, along with policies and programs that reduce these burdens for families with socioeconomic disadvantage

    Deconstruction of compound objects from image sets

    Full text link
    We propose a method to recover the structure of a compound object from multiple silhouettes. Structure is expressed as a collection of 3D primitives chosen from a pre-defined library, each with an associated pose. This has several advantages over a volume or mesh representation both for estimation and the utility of the recovered model. The main challenge in recovering such a model is the combinatorial number of possible arrangements of parts. We address this issue by exploiting the sparse nature of the problem, and show that our method scales to objects constructed from large libraries of parts

    Online Metric-Weighted Linear Representations for Robust Visual Tracking

    Full text link
    In this paper, we propose a visual tracker based on a metric-weighted linear representation of appearance. In order to capture the interdependence of different feature dimensions, we develop two online distance metric learning methods using proximity comparison information and structured output learning. The learned metric is then incorporated into a linear representation of appearance. We show that online distance metric learning significantly improves the robustness of the tracker, especially on those sequences exhibiting drastic appearance changes. In order to bound growth in the number of training samples, we design a time-weighted reservoir sampling method. Moreover, we enable our tracker to automatically perform object identification during the process of object tracking, by introducing a collection of static template samples belonging to several object classes of interest. Object identification results for an entire video sequence are achieved by systematically combining the tracking information and visual recognition at each frame. Experimental results on challenging video sequences demonstrate the effectiveness of the method for both inter-frame tracking and object identification.Comment: 51 pages. Appearing in IEEE Transactions on Pattern Analysis and Machine Intelligenc

    FVQA: Fact-based Visual Question Answering

    Full text link
    Visual Question Answering (VQA) has attracted a lot of attention in both Computer Vision and Natural Language Processing communities, not least because it offers insight into the relationships between two important sources of information. Current datasets, and the models built upon them, have focused on questions which are answerable by direct analysis of the question and image alone. The set of such questions that require no external information to answer is interesting, but very limited. It excludes questions which require common sense, or basic factual knowledge to answer, for example. Here we introduce FVQA, a VQA dataset which requires, and supports, much deeper reasoning. FVQA only contains questions which require external information to answer. We thus extend a conventional visual question answering dataset, which contains image-question-answerg triplets, through additional image-question-answer-supporting fact tuples. The supporting fact is represented as a structural triplet, such as . We evaluate several baseline models on the FVQA dataset, and describe a novel model which is capable of reasoning about an image on the basis of supporting facts.Comment: 16 page

    Ask Me Anything: Free-form Visual Question Answering Based on Knowledge from External Sources

    Full text link
    We propose a method for visual question answering which combines an internal representation of the content of an image with information extracted from a general knowledge base to answer a broad range of image-based questions. This allows more complex questions to be answered using the predominant neural network-based approach than has previously been possible. It particularly allows questions to be asked about the contents of an image, even when the image itself does not contain the whole answer. The method constructs a textual representation of the semantic content of an image, and merges it with textual information sourced from a knowledge base, to develop a deeper understanding of the scene viewed. Priming a recurrent neural network with this combined information, and the submitted question, leads to a very flexible visual question answering approach. We are specifically able to answer questions posed in natural language, that refer to information not contained in the image. We demonstrate the effectiveness of our model on two publicly available datasets, Toronto COCO-QA and MS COCO-VQA and show that it produces the best reported results in both cases.Comment: Accepted to IEEE Conf. Computer Vision and Pattern Recognitio

    Stuck in the moment: cognitive inflexibility in preschoolers following an extended time period

    Get PDF
    Preschoolers display surprising inflexibility in problem solving, but seem to approach new challenges with a fresh slate. We provide evidence that while the former is true the latter is not. Here, we examined whether brief exposure to stimuli can influence children’s problem solving following several weeks after first exposure to the stimuli. We administered a common executive function task, the Dimensional Change Card Sort, which requires children to sort picture cards by one dimension (e.g., color) and then switch to sort the same cards by a conflicting dimension (e.g., shape). After a week or after a month delay, we administered the second rule again. We found that 70% of preschoolers continued to sort by the initial sorting rule, even after a month delay, and even though they are explicitly told what to do. We discuss implications for theories of executive function development, and for classroom learning

    What value do explicit high level concepts have in vision to language problems?

    Full text link
    Much of the recent progress in Vision-to-Language (V2L) problems has been achieved through a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). This approach does not explicitly represent high-level semantic concepts, but rather seeks to progress directly from image features to text. We propose here a method of incorporating high-level concepts into the very successful CNN-RNN approach, and show that it achieves a significant improvement on the state-of-the-art performance in both image captioning and visual question answering. We also show that the same mechanism can be used to introduce external semantic information and that doing so further improves performance. In doing so we provide an analysis of the value of high level semantic information in V2L problems.Comment: Accepted to IEEE Conf. Computer Vision and Pattern Recognition 2016. Fixed titl
    corecore